30 research outputs found

    The marginalization paradox and the formal Bayes' law

    Full text link
    It has recently been shown that the marginalization paradox (MP) can be resolved by interpreting improper inferences as probability limits. The key to the resolution is that probability limits need not satisfy the formal Bayes' law, which is used in the MP to deduce an inconsistency. In this paper, I explore the differences between probability limits and the more familiar pointwise limits, which do imply the formal Bayes' law, and show how these differences underlie some key differences in the interpretation of the MP.Comment: Presented at Maxent 2007, Saratoga Springs, NY, July 200

    Information and Entropy

    Full text link
    What is information? Is it physical? We argue that in a Bayesian theory the notion of information must be defined in terms of its effects on the beliefs of rational agents. Information is whatever constrains rational beliefs and therefore it is the force that induces us to change our minds. This problem of updating from a prior to a posterior probability distribution is tackled through an eliminative induction process that singles out the logarithmic relative entropy as the unique tool for inference. The resulting method of Maximum relative Entropy (ME), which is designed for updating from arbitrary priors given information in the form of arbitrary constraints, includes as special cases both MaxEnt (which allows arbitrary constraints) and Bayes' rule (which allows arbitrary priors). Thus, ME unifies the two themes of these workshops -- the Maximum Entropy and the Bayesian methods -- into a single general inference scheme that allows us to handle problems that lie beyond the reach of either of the two methods separately. I conclude with a couple of simple illustrative examples.Comment: Presented at MaxEnt 2007, the 27th International Workshop on Bayesian Inference and Maximum Entropy Methods (July 8-13, 2007, Saratoga Springs, New York, USA

    Origins of the Combinatorial Basis of Entropy

    Full text link
    The combinatorial basis of entropy, given by Boltzmann, can be written H=N1lnWH = N^{-1} \ln \mathbb{W}, where HH is the dimensionless entropy, NN is the number of entities and W\mathbb{W} is number of ways in which a given realization of a system can occur (its statistical weight). This can be broadened to give generalized combinatorial (or probabilistic) definitions of entropy and cross-entropy: H=κ(ϕ(W)+C)H=\kappa (\phi(\mathbb{W}) +C) and D=κ(ϕ(P)+C)D=-\kappa (\phi(\mathbb{P}) +C), where P\mathbb{P} is the probability of a given realization, ϕ\phi is a convenient transformation function, κ\kappa is a scaling parameter and CC an arbitrary constant. If W\mathbb{W} or P\mathbb{P} satisfy the multinomial weight or distribution, then using ϕ()=ln()\phi(\cdot)=\ln(\cdot) and κ=N1\kappa=N^{-1}, HH and DD asymptotically converge to the Shannon and Kullback-Leibler functions. In general, however, W\mathbb{W} or P\mathbb{P} need not be multinomial, nor may they approach an asymptotic limit. In such cases, the entropy or cross-entropy function can be {\it defined} so that its extremization ("MaxEnt'' or "MinXEnt"), subject to the constraints, gives the ``most probable'' (``MaxProb'') realization of the system. This gives a probabilistic basis for MaxEnt and MinXEnt, independent of any information-theoretic justification. This work examines the origins of the governing distribution P\mathbb{P}.... (truncated)Comment: MaxEnt07 manuscript, version 4 revise

    From Information Geometry to Newtonian Dynamics

    Get PDF
    Newtonian dynamics is derived from prior information codified into an appropriate statistical model. The basic assumption is that there is an irreducible uncertainty in the location of particles so that the state of a particle is defined by a probability distribution. The corresponding configuration space is a statistical manifold the geometry of which is defined by the information metric. The trajectory follows from a principle of inference, the method of Maximum Entropy. No additional "physical" postulates such as an equation of motion, or an action principle, nor the concepts of momentum and of phase space, not even the notion of time, need to be postulated. The resulting entropic dynamics reproduces the Newtonian dynamics of any number of particles interacting among themselves and with external fields. Both the mass of the particles and their interactions are explained as a consequence of the underlying statistical manifold.Comment: Presented at MaxEnt 2007, the 27th International Workshop on Bayesian Inference and Maximum Entropy Methods (July 8-13, 2007, Saratoga Springs, New York, USA

    Updating Probabilities with Data and Moments

    Full text link
    We use the method of Maximum (relative) Entropy to process information in the form of observed data and moment constraints. The generic "canonical" form of the posterior distribution for the problem of simultaneous updating with data and moments is obtained. We discuss the general problem of non-commuting constraints, when they should be processed sequentially and when simultaneously. As an illustration, the multinomial example of die tosses is solved in detail for two superficially similar but actually very different problems.Comment: Presented at the 27th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Saratoga Springs, NY, July 8-13, 2007. 10 pages, 1 figure V2 has a small typo in the end of the appendix that was fixed. aj=mj+1 is now aj=m(k-j)+

    Information Geometry and Chaos on Negatively Curved Statistical Manifolds

    Full text link
    A novel information-geometric approach to chaotic dynamics on curved statistical manifolds based on Entropic Dynamics (ED) is suggested. Furthermore, an information-geometric analogue of the Zurek-Paz quantum chaos criterion is proposed. It is shown that the hyperbolicity of a non-maximally symmetric 6N-dimensional statistical manifold M_{s} underlying an ED Gaussian model describing an arbitrary system of 3N non-interacting degrees of freedom leads to linear information-geometric entropy growth and to exponential divergence of the Jacobi vector field intensity, quantum and classical features of chaos respectively.Comment: 7 pages, presented at MaxEnt2007, the 27th International Workshop on Bayesian Inference and Maximum Entropy Methods, Saratoga (NY, USA) (July-2007

    Lessons about likelihood functions from nuclear physics

    Get PDF
    Least-squares data analysis is based on the assumption that the normal (Gaussian) distribution appropriately characterizes the likelihood, that is, the conditional probability of each measurement d, given a measured quantity y, p(d | y). On the other hand, there is ample evidence in nuclear physics of significant disagreements among measurements, which are inconsistent with the normal distribution, given their stated uncertainties. In this study the histories of 99 measurements of the lifetimes of five elementary particles are examined to determine what can be inferred about the distribution of their values relative to their stated uncertainties. Taken as a whole, the variations in the data are somewhat larger than their quoted uncertainties would indicate. These data strongly support using a Student t distribution for the likelihood function instead of a normal. The most probable value for the order of the t distribution is 2.6 +/- 0.9. It is shown that analyses based on long-tailed t-distribution likelihoods gracefully cope with outlying data.Comment: presented at 27th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering (Maxent 2007), 10 pages, 12 figure

    On Shannon-Jaynes Entropy and Fisher Information

    Full text link
    The fundamentals of the Maximum Entropy principle as a rule for assigning and updating probabilities are revisited. The Shannon-Jaynes relative entropy is vindicated as the optimal criterion for use with an updating rule. A constructive rule is justified which assigns the probabilities least sensitive to coarse-graining. The implications of these developments for interpreting physics laws as rules of inference upon incomplete information are briefly discussed.Comment: Presented at the 27th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Saratoga Springs, NY, July 8-13, 200
    corecore